Stochastic gradient descent

(Stochastic gradient descent (SGD))

Efficient offline optimization method for functions ff with finite sum structure: f(𝐱)=i=1nfi(𝐱)f(\mathbf{x})=\sum_{i=1}^n f_i(\mathbf{x}) Goal is to find 𝐱̂\hat{\mathbf{x}} such that


Gradient descent

#incomplete


References: 1. 2. https://web.stanford.edu/class/ee270/scribes/lecture16.pdf